Task-specific dependency-based word embedding methods
نویسندگان
چکیده
While most traditional word embedding methods target generic tasks, two task-specific dependency-based are proposed for better performance in text classification tasks this work. First, we exploit the dependency parsing tree structure to capture structural information of a sentence, and develop method called (DWE). It finds keywords neighbor words as contexts via parsing. Next, leverage word-class co-occurrence statistics model class distributional incorporate it into learning process. This leads class-enhanced (CEDWE) method. Task-specific corpora matrix-factorization-based framework used train DWE CEDWE. Seven datasets evaluate CEDWE, experimental results show that they outperform several state-of-the-art methods.
منابع مشابه
An Improved Crowdsourcing Based Evaluation Technique for Word Embedding Methods
In this proposal track paper, we have presented a crowdsourcing-based word embedding evaluation technique that will be more reliable and linguistically justified. The method is designed for intrinsic evaluation and extends the approach proposed in (Schnabel et al., 2015). Our improved evaluation technique captures word relatedness based on the word context.
متن کاملDependency-Based Word Embeddings
While continuous word embeddings are gaining popularity, current models are based solely on linear contexts. In this work, we generalize the skip-gram model with negative sampling introduced by Mikolov et al. to include arbitrary contexts. In particular, we perform experiments with dependency-based contexts, and show that they produce markedly different embeddings. The dependencybased embedding...
متن کاملQLUT at SemEval-2017 Task 2: Word Similarity Based on Word Embedding and Knowledge Base
This paper shows the details of our system submissions in the task 2 of SemEval 2017. We take part in the subtask 1 of this task, which is an English monolingual subtask. This task is designed to evaluate the semantic word similarity of two linguistic items. The results of runs are assessed by standard Pearson and Spearman correlation, contrast with official gold standard set. The best performa...
متن کاملClass-specific Word Embedding through Linear Compositionality
English linguist John Rupert Firth has a famous saying “you shall know a word by the company it keeps.” Most word representation learning models are based on this assumption that a word’s semantic meaning can be learned from the context in which it resides. The context is defined as a small unordered number of words surrounding the target word. Research has shown that context alone provides lim...
متن کاملContradiction Detection with Contradiction-Specific Word Embedding
Contradiction detection is a task to recognize contradiction relations between a pair of sentences. Despite the effectiveness of traditional context-based word embedding learning algorithms in many natural language processing tasks, such algorithms are not powerful enough for contradiction detection. Contrasting words such as “overfull” and “empty” are mostly mapped into close vectors in such e...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Pattern Recognition Letters
سال: 2022
ISSN: ['1872-7344', '0167-8655']
DOI: https://doi.org/10.1016/j.patrec.2022.05.016